Periodic Stepsize Adaptation

نویسندگان

  • Chun-Nan Hsu
  • Han-Shen Huang
  • Yu-Ming Chang
چکیده

Previously, Bottou and LeCun [1] established that the second-order stochastic gradient descent (SGD) method can potentially achieve generalization performance as well as empirical optimum in a single pass through the training examples. However, second-order SGD requires computing the inverse of the Hessian matrix of the loss function, which is usually prohibitively expensive. Recently, we invented a new second-order SGD method, called Periodic Stepsize Adaptation (PSA). PSA explores a simple linear relation between the Hessian matrix and the Jacobian matrix of the mapping function. Instead of approximating Hessian, PSA approximates the Jacobian matrix which is proved to be simpler and more effective than approximating Hessian in an on-line setting. Experimental results for conditional random fields (CRF) and neural networks (NN) show that single-pass performance of PSA is very close to empirical optimum.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

L4: Practical loss-based stepsize adaptation for deep learning

We propose a stepsize adaptation scheme for stochastic gradient descent. It operates directly with the loss function and rescales the gradient in order to make fixed predicted progress on the loss. We demonstrate its capabilities by strongly improving the performance of Adam and Momentum optimizers. The enhanced optimizers with default hyperparameters consistently outperform their constant step...

متن کامل

New Proposed Second-Order ASDM Using OTAs

In this work, we present a new proposal for the second-order Adaptive Sigma Delta Modulation (ASDM). The new proposed Adaptation scheme is based on using Operational Transconductance Amplifier (OTAs) as an Integrator and as an Amplifier to adapt the quantizer stepsize to control the voltage gain by feedback the quantizer output through adaptation scheme. The stepsize is changing up or down by t...

متن کامل

Recursive Adaptation of Stepsize Parameter for Non-stationary Environments

In this article, we propose a method to adapt stepsize parameters used in reinforcement learning for dynamic environments. In general reinforcement learning situations, a stepsize parameter is decreased to zero during learning, because the environment is generally supposed to be noisy but stationary, such that the true expected rewards are fixed. On the other hand, we assume that in the real wo...

متن کامل

On the periodic solutions of discrete Hamiltonian systems

Almost all numerical methods for solving conservative problems cannot avoid a more or less perceptible drift phenomenon. Considering that the drift would be absent on a periodic or quasi-periodic solution, one way to eliminate such unpleasant phenomenon is to look for discrete periodic or quasi-periodic solutions. It is quite easy to show that only symmetric methods are able to provide solution...

متن کامل

An adaptive homotopy approach for non-selfadjoint eigenvalue problems

This paper presents adaptive algorithms for eigenvalue problems associated with non-selfadjoint partial differential operators. The basis for the developed algorithms is a homotopy method which departs from a wellunderstood selfadjoint problem. Apart from the adaptive grid refinement, the progress of the homotopy as well as the solution of the iterative method are adapted to balance the contrib...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2008